# Japanese enhancement

Gemma 2 Llama Swallow 9b It V0.1
The Gemma-2-Llama-Swallow series of models are multilingual large models constructed through continuous pre-training based on Gemma-2, with a particular enhancement in Japanese ability.
Large Language Model Transformers Supports Multiple Languages
G
tokyotech-llm
2,491
3
Gemma2 Alicia 9B
A multilingual Gemma-2 model fused using the TIES method, integrating expertise from multiple specialized domains
Large Language Model Transformers Supports Multiple Languages
G
yamatazen
15
2
Llama 3.3 Swallow 70B V0.4
Llama 3.3 Swallow is a large language model with 70 billion parameters, built on Meta Llama 3.3, which enhances Japanese capabilities while retaining English capabilities.
Large Language Model Transformers Supports Multiple Languages
L
tokyotech-llm
1,950
3
Llama 3.1 Swallow 70B Instruct V0.3
Llama 3.1 Swallow is a series of large language models built on Meta Llama 3.1. It enhances Japanese language capabilities through continuous pre-training while retaining English language capabilities.
Large Language Model Transformers Supports Multiple Languages
L
tokyotech-llm
1,659
12
Llama 3.1 Swallow 8B Instruct V0.3
Llama 3.1 Swallow is a series of large language models built on Meta Llama 3.1. It enhances Japanese capabilities through continuous pre-training while retaining English capabilities.
Large Language Model Transformers Supports Multiple Languages
L
tokyotech-llm
16.48k
20
Llama 3.1 Swallow 8B Instruct V0.2
Llama 3.1 Swallow is a series of large language models that are continuously pre - trained based on the Meta Llama 3.1 model, enhancing Japanese capabilities while retaining English capabilities.
Large Language Model Transformers Supports Multiple Languages
L
tokyotech-llm
2,283
15
Swallow MX 8x7b NVE V0.1
Apache-2.0
Swallow-MX-8x7b-NVE-v0.1 is a Mixture of Experts model based on Mixtral-8x7B-Instruct-v0.1 with continued pretraining, primarily enhancing Japanese capabilities.
Large Language Model Transformers Supports Multiple Languages
S
tokyotech-llm
1,293
29
Karakuri Lm 70b Chat V0.1
Other
KARAKURI LM is a pre-trained language model built on Llama 2, which enhances Japanese processing capabilities and is further pre-trained on Japanese and multilingual corpora.
Large Language Model Transformers Supports Multiple Languages
K
karakuri-ai
2,300
24
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase